With the growth of editing and sharing images through the internet, the importance of protecting the images' authorship has increased. Robust watermarking is a known approach to maintaining copyright protection. Robustness and imperceptibility are two factors that are tried to be maximized through watermarking. Usually, there is a trade-off between these two parameters. Increasing the robustness would lessen the imperceptibility of the watermarking. This paper proposes an adaptive method that determines the strength of the watermark embedding in different parts of the cover image regarding its texture and brightness. Adaptive embedding increases the robustness while preserving the quality of the watermarked image. Experimental results also show that the proposed method can effectively reconstruct the embedded payload in different kinds of common watermarking attacks. Our proposed method has shown good performance compared to a recent technique.
translated by 谷歌翻译
Recently, many attempts have been made to construct a transformer base U-shaped architecture, and new methods have been proposed that outperformed CNN-based rivals. However, serious problems such as blockiness and cropped edges in predicted masks remain because of transformers' patch partitioning operations. In this work, we propose a new U-shaped architecture for medical image segmentation with the help of the newly introduced focal modulation mechanism. The proposed architecture has asymmetric depths for the encoder and decoder. Due to the ability of the focal module to aggregate local and global features, our model could simultaneously benefit the wide receptive field of transformers and local viewing of CNNs. This helps the proposed method balance the local and global feature usage to outperform one of the most powerful transformer-based U-shaped models called Swin-UNet. We achieved a 1.68% higher DICE score and a 0.89 better HD metric on the Synapse dataset. Also, with extremely limited data, we had a 4.25% higher DICE score on the NeoPolyp dataset. Our implementations are available at: https://github.com/givkashi/Focal-UNet
translated by 谷歌翻译
图像重新定位旨在更改图像大小,同时保留重要内容并最大程度地减少明显的扭曲。但是,先前的图像重新定位方法创建了遭受工件和扭曲的输出。此外,大多数以前的作品都尝试同时重新定位输入图像的背景和前景。同时调整前景和背景会导致对象的长宽比的变化。纵横比的变化对于人类对象并不理想。我们提出了一种克服这些问题的重新定位方法。提出的方法包括以下步骤。首先,一种涂上方法使用输入图像和前景对象的二进制掩码来生成背景图像,而无需任何前景对象。其次,接缝雕刻方法将背景图像调整到目标大小。然后,一种超分辨率方法增加了输入图像质量,然后提取前景对象。最后,将重定位的背景和提取的超级分辨对象馈入粒子群优化算法(PSO)中。 PSO算法使用审美质量评估作为其目标函数,以确定将对象放置在背景中的最佳位置和大小。我们使用图像质量评估和美学质量评估措施来显示我们与流行的图像重新定位技术相比的优越结果。
translated by 谷歌翻译
MRI图像中的脑肿瘤分析是一个重要而挑战性的问题,因为误诊可能导致死亡。脑肿瘤在早期阶段的诊断和评估增加了成功治疗的概率。然而,肿瘤,形状和位置的复杂性和各种使其分割和分类复合物。在这方面,许多研究人员提出了脑肿瘤细分和分类方法。本文使用含有MRI图像增强和肿瘤区检测的框架,呈现了一种同时分段和分类MRI图像中的脑肿瘤的方法。最终,提出了一种基于多任务学习方法的网络。主观和客观结果表明,基于评估指标的分割和分类结果更好或与最先进的。
translated by 谷歌翻译
图像染色是增强扭曲数字图像的有效方法。不同的初始化方法使用相邻像素的信息来预测丢失像素的值。最近,深度神经网络已经用于学习图像的结构和语义细节以获得避免目的。在本文中,我们提出了一种用于图像染色的网络。此网络类似于U-Net,从图像中提取各种功能,导致更好的结果。我们通过用输出图像的恢复像素替换损坏的像素来改善最终结果。我们的实验结果表明,该方法产生了与传统方法相比的高质量结果。
translated by 谷歌翻译
数字病理学是现代医学中最重要的发展之一。病理检查是医疗方案的黄金标准,并在诊断中发挥基本作用。最近,随着数字扫描仪的出现,现在可以将组织组织病理学载玻片数字化并作为数字图像存储。结果,数字化组织病理组织可用于计算机辅助图像分析程序和机器学习技术。核的检测和分割是癌症诊断中的一些基本步骤。最近,深度学习已被用于核细胞分割。然而,核细胞分割的深度学习方法中的一个问题是缺乏斑块的信息。本文提出了深入的基于学习的核细胞分割方法,这解决了补丁边界地区误入歧途的问题。我们使用本地和全局修补程序来预测最终的分割图。多器官组织病理学数据集上的实验结果表明,我们的方法优于基线核细胞分割和流行分割模型。
translated by 谷歌翻译
根据世界卫生组织(世卫组织),癌症是全世界第二次死亡原因,仅对2018年的950万人死亡负责。脑肿瘤计数每四个癌症死亡中的一次。因此,准确和及时诊断脑肿瘤会导致更有效的治疗方法。医生只通过脑手术进行活组织检查操作,并且在诊断肿瘤类型后,考虑治疗计划。基于机器学习算法的自动系统可以允许医生以非侵入性措施诊断脑肿瘤。迄今为止,已经提出了几种图像分类方法以辅助诊断和治疗。对于脑肿瘤分类在这项工作中,我们提供基于深度学习的系统,包含编码器块。这些块作为剩余学习的最大池特征送入。我们的方法展示了通过使用有限的医学图像数据集提高磁共振成像(MRI)图像中的肿瘤分类精度来实现有希望的结果。该模型在数据集中的实验评估由3064 MR图像组成的准确度提出95.98%,这比以前关于此数据库的研究更好。
translated by 谷歌翻译
Saliency detection is one of the most challenging problems in image analysis and computer vision. Many approaches propose different architectures based on the psychological and biological properties of the human visual attention system. However, there is still no abstract framework that summarizes the existing methods. In this paper, we offered a general framework for saliency models, which consists of five main steps: pre-processing, feature extraction, saliency map generation, saliency map combination, and post-processing. Also, we study different saliency models containing each level and compare their performance. This framework helps researchers to have a comprehensive view of studying new methods.
translated by 谷歌翻译
Semantic segmentation classifies each pixel in the image. Due to its advantages, semantic segmentation is used in many tasks, such as cancer detection, robot-assisted surgery, satellite image analysis, and self-driving cars. Accuracy and efficiency are the two crucial goals for this purpose, and several state-of-the-art neural networks exist. By employing different techniques, new solutions have been presented in each method to increase efficiency and accuracy and reduce costs. However, the diversity of the implemented approaches for semantic segmentation makes it difficult for researchers to achieve a comprehensive view of the field. In this paper, an abstraction model for semantic segmentation offers a comprehensive view of the field. The proposed framework consists of four general blocks that cover the operation of the majority of semantic segmentation methods. We also compare different approaches and analyze each of the four abstraction blocks' importance in each method's operation.
translated by 谷歌翻译